Learning Hidden Variables in Probabilistic Graphical Models

نویسندگان

  • Gal Elidan
  • Nir Friedman
چکیده

In the past decades, a great deal of research has focused on learning probabilistic graphical models from data. A serious problem in learning such models is the presence of hidden, or latent, variables. These variables are not observed, yet their interaction with the observed variables has important consequences in terms of representation, inference and prediction. Consequently, numerous works have been directed towards learning probabilistic graphical models with hidden variables. A significantly harder challenge is that of detecting new hidden variables and incorporating them into the network structure. Surprisingly, and despite the recognized importance of hidden variables both in social sciences and the learning community, this problem has received little attention. In this dissertation we explore the problem of learning new hidden variable in real-life domains. We present methods for coping with the different elements that this task encompasses: the detection of new hidden variables; determining the cardinality of new hidden variables; incorporating new hidden variables into learning model. In addition we also address the problem of local maxima that is common in many learning scenarios, and is particularly acute in the presence of hidden variables. We present simple and easy to implement methods that work when training data is relatively plentiful as well as a more elaborate framework that is suitable when the model is particularly complex and the data is sparse. We also consider methods specifically tailored at networks with continuous variables and the added challenges in this scenario. We evaluate all of our methods on both synthetic and real-life data. For the more elaborate methods, we put a particular emphasis on learning complex models with many hidden variables. We demonstrate significant improvement in quantitative prediction on unseen test samples when learning with hidden variables, reaffirming their importance in practice. We also demonstrate that models learned with our methods have hidden variables that are qualitatively appealing and shed light on the learned domain.

برای دانلود رایگان متن کامل این مقاله و بیش از 32 میلیون مقاله دیگر ابتدا ثبت نام کنید

ثبت نام

اگر عضو سایت هستید لطفا وارد حساب کاربری خود شوید

منابع مشابه

A Guide to the Literature on Learning Probabilistic Networks from Data

|This literature review discusses di erent methods under the general rubric of learning Bayesian networks from data, and includes some overlapping work on more general probabilistic networks. Connections are drawn between the statistical, neural network, and uncertainty communities, and between the di erentmethodological communities, such as Bayesian, description length, and classical statistic...

متن کامل

Discrete restricted Boltzmann machines

We describe discrete restricted Boltzmann machines: probabilistic graphical models with bipartite interactions between visible and hidden discrete variables. Examples are binary restricted Boltzmann machines and discrete näıve Bayes models. We detail the inference functions and distributed representations arising in these models in terms of configurations of projected products of simplices and ...

متن کامل

Buntine : a Guide to the Literature on Learning Graphical Models from Data

| This literature review discusses diierent methods under the general rubric of learning Bayesian networks from data, and includes some overlapping work on more general probabilistic networks. Connections are drawn between the statistical, neural network, and uncertainty communities , and between the diierent methodological communities, such as Bayesian, description length, and classical statis...

متن کامل

Probabilistic Language Modeling with Hidden Stochastic Automata

In this paper, we introduce a novel dynamical Bayesian network model for probabilistic language modeling. We refer to this as the Hidden Stochastic Automaton. This model, while based on a generalization of the Hidden Markov model, has qualitatively greater generative power than either the Hidden Markov model itself or any of its existing variants and generalizations. This allows the Hidden Stoc...

متن کامل

Build, Compute, Critique, Repeat: Data Analysis with Latent Variable Models

We survey how to use latent variable models to solve data analysis problems. A latent variable model is a probabilistic model of hidden and observed variables, where the hidden variables encode hidden patterns in our data. We uncover these patterns through the posterior, the conditional distribution of the hidden variables given the observations, which we use to explore, summarize, and form pre...

متن کامل

ذخیره در منابع من


  با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید

عنوان ژورنال:

دوره   شماره 

صفحات  -

تاریخ انتشار 2004